Improving Deep Neural Network with Multiple Parametric Exponential Linear Units
نویسندگان
چکیده
Activation function is crucial to the recent successes of neural network. In this paper, we propose a new activation function that generalizes and unifies the rectified and exponential linear units. The proposed method, named MPELU, has the advantages of PReLU and ELU. We show that by introducing learnable parameters like PReLU, MPELU provides better generalization capability than PReLU and ELU. As a generalization of ELU, MPELU maintains the ability of reducing bias shift and shows better convergence than ReLU and PReLU. In addition, we put forward a way of initialization suitable for exponential linear units. To the best of our knowledge, the proposed initialization is the first one for exponential-linear-unit networks. Experiments prove that our initialization is helpful to train very deep exponential-linear-unit networks. And exponential-linear-unit networks with the proposed initialization, to some extent, are able to deal with the degradation phenomenon caused by depth.
منابع مشابه
Improving Neural Networks with Dropout by Nitish Srivastava
Improving Neural Networks with Dropout Nitish Srivastava Master of Science Graduate Department of Computer Science University of Toronto 2013 Deep neural nets with a huge number of parameters are very powerful machine learning systems. However, overfitting is a serious problem in such networks. Large networks are also slow to use, making it difficult to deal with overfitting by combining many d...
متن کاملFlexible Rectified Linear Units for Improving Convolutional Neural Networks
Rectified linear unit (ReLU) is a widely used activation function for deep convolutional neural networks. In this paper, we propose a novel activation function called flexible rectified linear unit (FReLU). FReLU improves the flexibility of ReLU by a learnable rectified point. FReLU achieves a faster convergence and higher performance. Furthermore, FReLU does not rely on strict assumptions by s...
متن کاملINFLUENCE OF FIBER ASPECT RATIO ON SHEAR CAPACITY OF DEEP BEAMS USING ARTIFICIAL NEURAL NETWORK TECHNIQUE
This paper deals with the effect of fiber aspect ratio of steel fibers on shear strength of steel fiber reinforced concrete deep beams loaded with shear span to depth ratio less than two using the artificial neural network technique. The network model predicts reasonably good results when compared with the equation proposed by previous researchers. The parametric study invol...
متن کاملImproving Deep Neural Networks with Probabilistic Maxout Units
We present a probabilistic variant of the recently introduced maxout unit. The success of deep neural networks utilizing maxout can partly be attributed to favorable performance under dropout, when compared to rectified linear units. It however also depends on the fact that each maxout unit performs a pooling operation over a group of linear transformations and is thus partially invariant to ch...
متن کاملImproving deep convolutional neural networks with mixed maxout units
Motivated by insights from the maxout-units-based deep Convolutional Neural Network (CNN) that "non-maximal features are unable to deliver" and "feature mapping subspace pooling is insufficient," we present a novel mixed variant of the recently introduced maxout unit called a mixout unit. Specifically, we do so by calculating the exponential probabilities of feature mappings gained by applying ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1606.00305 شماره
صفحات -
تاریخ انتشار 2016